AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Single-GPU Efficient Inference

# Single-GPU Efficient Inference

Jamba V0.1 9B
Apache-2.0
Jamba is a state-of-the-art hybrid SSM-Transformer architecture large language model, combining the advantages of attention mechanisms and the Mamba architecture, supporting a 256K context length, and suitable for inference on a single 80GB GPU.
Large Language Model Transformers
J
TechxGenus
22
10
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase